Analysis of Parallelization Techniques and Tools
نویسندگان
چکیده
Parallel Computing solves computationally large problems by partitioning into multiple tasks and running simultaneously on multicore or multiprocessor environment based on shared or distributed memory architectures. New multicore era demands software programmer to develop parallel programs to completely utilize the hardware parallelism. Writing parallel program manually for complex problem is a tedious and time consuming process. Hence, automatic parallelization tools were evolved to automate the process of converting sequential code to parallel. Parallelization techniques like Dependency Analysis and Loop Parallelization plays a major role in parallelization process. This paper aims at providing brief survey of existing OpenMP based parallelization tools and also the anlaysis on the performance of two such tools on typical problems. The OpenMP directives based tools Cetus and Par4all are taken for our case studies. The performance of the automated parallelization on sample program were studied and analyzed in terms of execution time and speedup. The study revealed the conditions under which they could effectively parallelize the code and the conditions they could not handle as effectively as manual parallelization. Also many of the available parallelization tools are meant for shared memory hardware architecture and relatively a very few automated tools based on MPI are available for distributed memory architectures. The paper underscores the need for efficient parallelization tools supporting different parallel processing environments, with the ability to identify and exploit the parallelism by inserting parallel directives or APIs in serial programs. S. Prema & R. Jehadeesan 472
منابع مشابه
Parallelization of Rich Models for Steganalysis of Digital Images using a CUDA-based Approach
There are several different methods to make an efficient strategy for steganalysis of digital images. A very powerful method in this area is rich model consisting of a large number of diverse sub-models in both spatial and transform domain that should be utilized. However, the extraction of a various types of features from an image is so time consuming in some steps, especially for training pha...
متن کاملEecient Interprocedural Array Data--ow Analysis for Automatic Program Parallelization Accepted to Ieee Transactions on Software Engineering, the Special Issue on Architecture-independent Lan- Guages and Software Tools for Parallel Processing
Since sequential languages such as Fortran and C are more machine-independent than current parallel languages, it is highly desirable to develop powerful parallelization tools which can generate parallel codes, automatically or semi-automatically, targeting diierent parallel architectures. Array data-ow analysis is known to be crucial to the success of automatic parallelization. Such an analysi...
متن کاملAn Overview of OpenMP based Automatic Parallelization Tools
In the fast paced world today, computer programmers are required to know everything about speeding up and optimizing an application. Since the barriers for speeding up sequential code have been reached, we now have to find other options. And this is where parallel programs come into picture. A programmer thus has to learn how to parallelize code and would require a comprehensive understanding o...
متن کاملUtilizing Strategic Management Accounting Techniques in Iranian Firms
The main objective of this article is to describe and explain the utilization of Strategic Management Accounting Techniques (SMATs) in Iranian various firms. For this, a survey was carried out using questionnaires provided for the seventy-five Chief Executive Officers in productive and services firms. Data gathered from respondents about the usage rate of SMATs based on the five-point Likert sc...
متن کاملTransparent Parallelization of Binary Code
This paper describes a system that applies automatic parallelization techniques to binary code. The system works by raising raw executable code to an intermediate representation that exhibits all memory accesses and relevant register definitions, but outlines detailed computations that are not relevant for parallelization. It then uses an off-the-shelf polyhedral parallelizer, first applying ap...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2013